Representation Learning by Detecting Incorrect Location Embeddings
نویسندگان
چکیده
In this paper, we introduce a novel self-supervised learning (SSL) loss for image representation learning. There is growing belief that generalization in deep neural networks linked to their ability discriminate object shapes. Since shape related the location of its parts, propose detect those have been artificially misplaced. We represent parts with tokens and train ViT which token has combined an incorrect positional embedding. then sparsity inputs make model more robust occlusions speed up training. call our method DILEMMA, stands Detection Incorrect Location EMbeddings MAsked inputs. apply DILEMMA MoCoV3, DINO SimCLR show improvement performance respectively 4.41%, 3.97%, 0.5% under same training time linear probing transfer on ImageNet-1K. also full fine-tuning improvements MAE ImageNet-100. evaluate via common SSL benchmarks. Moreover, when downstream tasks are strongly reliant (such as YOGA-82 pose dataset), pre-trained features yield significant gain over prior work.
منابع مشابه
Detecting Incorrect Numerical Data in DBpedia
DBpedia is a central hub of Linked Open Data (LOD). Being based on crowd-sourced contents and heuristic extraction methods, it is not free of errors. In this paper, we study the application of unsupervised numerical outlier detection methods to DBpedia, using Interquantile Range (IQR), Kernel Density Estimation (KDE), and various dispersion estimators, combined with different semantic grouping ...
متن کاملMulticlass Learning by Probabilistic Embeddings
We describe a new algorithmic framework for learning multiclass categorization problems. In this framework a multiclass predictor is composed of a pair of embeddings that map both instances and labels into a common space. In this space each instance is assigned the label it is nearest to. We outline and analyze an algorithm, termed Bunching, for learning the pair of embeddings from labeled data...
متن کاملRepresentation Tradeoffs for Hyperbolic Embeddings
Hyperbolic embeddings offer excellent quality with few dimensions when embedding hierarchical data structures like synonym or type hierarchies. Given a tree, we give a combinatorial construction that embeds the tree in hyperbolic space with arbitrarily low distortion without using optimization. On WordNet, our combinatorial embedding obtains a mean-average-precision of 0.989 with only two dimen...
متن کاملDetecting protein sequence conservation via metric embeddings
MOTIVATION Comparing two protein databases is a fundamental task in biosequence annotation. Given two databases, one must find all pairs of proteins that align with high score under a biologically meaningful substitution score matrix, such as a BLOSUM matrix (Henikoff and Henikoff, 1992). Distance-based approaches to this problem map each peptide in the database to a point in a metric space, su...
متن کاملDoes Incorrect Guessing Impair Fact Learning?
Taking a test has been shown to produce enhanced retention of the retrieved information. On tests, however, students often encounter questions the answers for which they are unsure. Should they guess anyway, even if they are likely to answer incorrectly? Or are errors engrained, impairing subsequent learning of the correct answer? We sought to answer this question in 3 experiments. In Experimen...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2023
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v37i8.26160